Improving Deep Mutual Learning via Knowledge Distillation
نویسندگان
چکیده
Knowledge transfer has become very popular in recent years, and it is either based on a one-way method used with knowledge distillation or two-way implemented by deep mutual learning, while both of them adopt teacher–student paradigm. A more simple compact because only involves an untrained low-capacity student high-capacity teacher network the process. In contrast, requires training costs two low-cost capacities from scratch simultaneously to obtain better accuracy results for each network. this paper, we propose new approaches, namely full learning (FDDML) half (HDDML), improve convolutional neural performance. These approaches work three losses using variations existing architectures, experiments have been conducted public benchmark datasets. We test our some KT task methods, showing its performance over related methods.
منابع مشابه
Deep Mutual Learning
Model distillation is an effective and widely used technique to transfer knowledge from a teacher to a student network. The typical application is to transfer from a powerful large network or ensemble to a small network, that is better suited to low-memory or fast execution requirements. In this paper, we present a deep mutual learning (DML) strategy where, rather than one way transfer between ...
متن کاملImproving Stock Return Forecasting by Deep Learning Algorithm
Improving return forecasting is very important for both investors and researchers in financial markets. In this study we try to aim this object by two new methods. First, instead of using traditional variable, gold prices have been used as predictor and compare the results with Goyal's variables. Second, unlike previous researches new machine learning algorithm called Deep learning (DP) has bee...
متن کاملData-Free Knowledge Distillation for Deep Neural Networks
Recent advances in model compression have provided procedures for compressing large neural networks to a fraction of their original size while retaining most if not all of their accuracy. However, all of these approaches rely on access to the original training set, which might not always be possible if the network to be compressed was trained on a very large dataset, or on a dataset whose relea...
متن کاملInterpreting Deep Classifier by Visual Distillation of Dark Knowledge
Interpreting black box classifiers, such as deep networks, allows an analyst to validate a classifier before it is deployed in a high-stakes setting. A natural idea is to visualize the deep network’s representations, so as to “see what the network sees”. In this paper, we demonstrate that standard dimension reduction methods in this setting can yield uninformative or even misleading visualizati...
متن کاملImproving the fidelity of optical Zeno gates via distillation
We have modeled the Zeno effect controlled-sign gate of Franson et al. Phys. Rev. A 70, 062302 2004 and shown that high two-photon to one-photon absorption ratios, , are needed for high fidelity free-standing operation. Hence we instead employ this gate for cluster state fusion, where the requirement for is less restrictive. With the help of partially offline one-photon and two-photon distillat...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Applied sciences
سال: 2022
ISSN: ['2076-3417']
DOI: https://doi.org/10.3390/app12157916